Original Paper
Abstract
Background: Caregivers supporting individuals with Alzheimer disease and related dementias (AD/ADRD) frequently encounter prolonged emotional strain, psychological distress, and social isolation, yet their needs are largely overlooked in current technological and clinical interventions. The special routines and obligations of caregivers of individuals with AD/ADRD are frequently not well-suited to the many artificial intelligence–driven (AI-driven) mental health solutions that are currently available. This reveals a critical need for sophisticated, customized solutions created especially to help the mental health of caregivers for patients with AD/ADRD.
Objective: To address the existing limitations of personalized mental health interventions, we aimed to identify existing literature on personalized mental health interventions using AI for specific purposes and to develop a new framework for the caregivers of individuals with AD/ADRD.
Methods: We followed an iterative approach to design the new framework. First, we did a systematic literature review of current literature to identify data analysis, AI methods, and personalized interventions. Second, we focused on the underlying gaps of this research, and by synthesizing our findings from the review, we proposed a conceptual framework.
Results: The systematic literature review identified 73 unique results, and from external sources, we found 3 unique potential papers. Of these, 28 papers were eligible for inclusion, on which we performed our analysis. Based on the findings, we developed a new conceptual framework with 3 special features that are specifically for caregivers of patients with AD/ADRD. The 3 unique features are a personalized daily routine scheduler, which will take both patients with AD/ADRD and caregiver’s information to make it personalized, a daily reward system to keep patients motivated, and an educational repository to get the bite-sized knowledge for the lesson of handling patients in an efficient manner and taking care of one’s own mental health.
Conclusions: The proposed framework provides a chance for caregivers to receive mental health care, which will be personalized. The framework is developed with more updated methods than existing approaches, with a lack of personalization in this sector. This framework can be implemented with a goal of personalization and explainable approaches and can undergo further iterations to ensure it is appropriate for specific purposes.
doi:10.2196/79973
Keywords
Introduction
Alzheimer disease and related dementias (AD/ADRD) demand care to give the patients full-time support, and in that case, formal or informal caregivers play a very important role. However, the demand for caregivers of individuals with AD/ADRD is currently surging with the older population who were born between 1946 and 1964 (also known as the baby boomer generation []). For instance, according to the Alzheimer Association [], in 2024, there were approximately 7 million Americans aged 65 years and older experiencing AD, and the number of patients might double by 2060. Therefore, this baby boomer generation is placing a high demand on caregivers as they are entering the high-risk age bracket for AD. This sudden need creates a critical shortage of formal caregivers and emotional challenges among informal caregivers, such as family members or loved ones.
Though the demand for caregivers is increasing rapidly, their number has not increased accordingly, burdening them with workloads beyond their capacity. This scarcity might put pressure on caregivers, especially informal caregivers who often bear the burden quietly in this situation. For instance, in the United States alone, more than 11 million unpaid caregivers offer 18 billion hours of care, worth nearly US $340 billion yearly, to patients with AD/ADRD, according to the Alzheimer Association []. The study by Sallim et al [] found that roughly 34% of caregivers experience depression, while 43.6% report anxiety. Apart from that, caregivers mostly have to deal with sedentary lifestyles, sleep deprivation, social isolation, and foretaste grief, which are not even treated by professionals due to time and financial constraints []. Therefore, there is a pressing need for a change in attention towards the mental health support system customized for caregivers of individuals with AD/ADRD, which will allow them to manage their burden and have some time for themselves to work on their well-being.
Caregivers of individuals with AD/ADRD need a system that will be personalized depending on the stage of patients with AD/ADRD, the caregiver’s own preferences, and background. There are various digitally accessible mental health supports for caregivers of individuals with AD/ADRD, such as informational portals and educational resources [,], therapy and counseling services [], peer-led support groups [,], and some digital tools and mobile apps [-]. The purpose of these interventions is to improve the mental health of caregivers for patients with AD/ADRD. However, existing interventions have some limitations, too. For instance, some mobile apps offer limited interactivity, minimal emotional support integration, and static content delivery, which lacks adaptive learning or personalization based on caregiver experience level or stress indicators [,]. Besides, some interventions [], require smart home infrastructure, which may not be accessible to low-income or older caregivers.
In order to provide more background of our study, we group the technologies discussed in this paper into six components: (1) natural language processing (NLP): analyzes text (eg, notes, messages, and forum posts) to detect topics, symptoms, or risk; (2) recommender systems and personalization policies: select the “next best” tip, task, or content and adjust timing based on user state and preferences; (3) conversational agents: rule-based or large language model chatbots that deliver check-ins or guidance; (4) multimodal sensing: passive phone/wearable signals (eg, sleep, steps, and heart rate) optionally combined with brief ecological momentary assessments (EMA); (5) modeling approaches: classical machine learning such as logistic regression, random forests and deep learning, for example, recurrent/transformer networks used for prediction; and (6) explainability and deployment practices: methods that show why a model made a suggestion such as feature importance, saliency and operational safeguards (calibration, drift monitoring, and privacy). We use this taxonomy throughout the paper, such as research question (RQ) 1 (RQ1) addresses data and processing, RQ2 addresses personalization and monitoring, and RQ3 addresses explainability and clinical integration.
To make a system more acceptable to users, certain artificial intelligence (AI) design guidelines should be followed []. Amershi et al [] presented 18 guidelines to ensure AI-infused apps are understandable, trustworthy, fair, and usable. Among the guidelines, guideline G11 is about making clear why the system did what it did. These explanations are highly important for users to trust the system and have transparency between users and the system. Though this guideline is crucial in creating human-centric AI systems, most of the recent caregiver apps [-] do not follow the guideline of explainable AI. Moreover, according to Jung et al [], the aspect of explainability is vital for fostering trust and ensuring that both caregivers and clinicians can comprehend and effectively apply AI-generated recommendations.
Rather than proceeding directly to the design of a caregiver-centered intervention, our study begins with a systematic review of personalized mental health technologies to establish a comprehensive understanding of current trends, limitations, and opportunities. This approach allows us to learn from and build upon the broader mental health innovation landscape before narrowing our focus to caregiving for individuals with AD/ADRD. The overview of this study is shown in and . To guide this review, we developed the following RQs:
- RQ1: What types of data are used for detecting early warning signs of mental disorders, and how are these data collected and processed?
- RQ2: What types of AI models are used in current depression monitoring systems, and what limitations hinder their effectiveness in real-world practice?
- RQ3: What usability and design challenges arise in implementing explainable human-AI interaction systems for mental health care?


By systematically reviewing these areas, we aim to extract key insights that will guide the design of our caregiver-focused mental health intervention framework. In this study, we have discussed the findings of the review in the Results section. Briefly, we found heterogeneous types of data sources, and among them, clinical trials and mobile apps were the most adapted. Besides, real-time monitoring was highly rare and even poorly reported. Personalization and explainability were present in some of the studies. However, those were inconsistently defined and rarely transparent. To summarize the findings, the identified gaps indicate the need for hybrid sensing, such as passive data augmented with brief self-reports, policy-transparent personalization, and clinician-oriented explanations to enable trustworthy adoption. Based on our findings, we are proposing a conceptual framework for personalized AI-driven mental health intervention. This framework is designed to support dynamic treatment personalization, explainability, and real-time user engagement, particularly for caregivers of Alzheimer-affected individuals. We derived the caregiver-specific components from proven dementia-caregiver interventions (psychoeducation or skills training, stress reduction, counseling, or peer support) and standard outcome frameworks (burden, mood, coping, and role functioning) [-]. We then translated these needs into low-burden digital features such as hybrid sensing with brief check-ins, just-in-time support, guided by Human-AI Interaction and Just-in-Time Adaptive Interventions principles [,,]. The framework is broadly discussed in the discussion section.
Methods
Database Selection and Search Strategy
In this systematic review, we followed the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines (). The review was started in February 2025 using the 5 main databases, such as PubMed, Google Scholar, IEEE Xplore, ScienceDirect, and Scopus, following specific search queries (provided in ). Only peer-reviewed publications published between January 2020 and February 2025 were included in the search parameters. To maintain methodological integrity, we have only included original research articles that were written in English language. In this review, book chapters, structured or unstructured reviews, observational notes, and position papers were not included.
The search string () was carefully designed to fit the emphasis of our work. Interventions of our interest included the use of smart devices and AI-enabled or precision medicine approaches for mental health. Moreover, we considered research indicating promise for personalized treatment. The study excluded studies that are not connected to mental health treatment or AI-assisted solutions.
Study Selection
One team member (SU) performed the initial screening of studies by title. Then, 2 individual reviewers (SU) and (CR) were given the task of carrying out the screening, determining eligibility, and managing the inclusion procedure. After careful consideration, any discrepancies between the 2 teams were fixed through agreement. Both reviewers independently then screened the potential studies by title and abstract. A third team (IS) was consulted to make the final decision in cases where an agreement could not be reached. From February to March 20, 2025, the screening procedure was carried out following the inclusion and exclusion criteria. Studies that did not relate to AI-driven personalized interventions for mental health treatment were not considered. Additionally, articles that focused on other medical conditions, were not subject to peer review, or were editorials, interviews, or position papers, were not included.
Selection and Inclusion of Studies
The PRISMA flow diagram summarizes the study selection process. The initial search yielded a total of 76 citations, from which 13 duplicates were removed. The remaining records were screened based on their titles and abstracts, leading to the exclusion of 15 citations that did not meet the inclusion criteria. The remaining 48 studies were then subjected to full-text review, during which 3 papers were excluded. The primary reasons for exclusion included the use of ineligible manuscript types such as published abstracts, studies that focused on mental health treatment without addressing personalized care or precision medicine, and papers that did not involve the application of machine learning or AI. After the screening process, a total of 28 studies were included in the final review.
Data Extraction
For eligible studies, data were extracted from the studies by 2 teams (IS and CR), and the whole process was reviewed by a third team (SU). Data were extracted from the potential studies by following the RQs. First, information about data sources, data collection methods, data types, preprocessing techniques, and AI models used in the studies was extracted from the reviewed studies. Second, both teams (IS and CR) gathered information about AI methodologies and approaches, real-time monitoring scopes, personalization strategies, intervention models, critical influencing factors, challenges in AI implementations, target population and user groups, and associated mental health conditions. Third, insights were gathered on AI explanation methods (including explainability and interpretability techniques), system design and implementation challenges, trade-offs between model performance and interpretability, clinician-centered barriers and real-world integrations, as well as the impact of AI on clinical decision-making and patient outcomes. After every step, the analysis was evaluated and fine-tuned by another team (SU).
Developing a Conceptual Framework
The team members of this research analyzed the findings of the review. During our analysis, we emphasized more on real time monitoring tools, personalized approaches, and AI models used. This thorough analysis helped us to develop our comprehensive conceptual framework.
Results
Overview
We analyzed the included studies () and organized the core findings based on RQs. The findings for our RQ1, RQ2, and RQ3 can be found in - [-], consequently. In addition, the shortcomings of single approaches and what complementary methods can be adopted in this case are shown in .

| Single approach | What it captures well | What it misses | Common failure modes | Complementary inputs/interventions | Why it fills the gap | Implementation note |
| Self-report only (mood check-ins, brief surveys) [,] | Subjective state, intent, symptom nuance | Day-to-day fluctuations, noncompliance periods, social desirability bias | Sparse/late entries, recall bias [,] | Passive sensing (sleep/steps/HRa), app usage logs [] | Objective, continuous context; flags when self-report is absent | Use opt-in sensing; prompt when passives drift from reported mood |
| Passive wearable only (HRVb, sleep, steps) [,,,,,] | Physiology, routines, nocturnal patterns | Cognition, emotion triggers, caregiver burden context | Sensor dropouts, artifacts, poor specificity [,] | Just-in-time self-report, short EMAc, text notes [] | Labels events/feelings that physiology cannot disambiguate | Micro-prompts (≤10 s) after anomalous windows |
| Smartphone/app logs only (usage, taps, screen time) [,,] | Engagement, adherence, micro-routines | Clinical severity, affect, reasons for lapses | “Silent churn”; high noise from noncare tasks [,] | Brief affect scale, care task checklist, physiology [] | Links behavior to state; separates care versus noncare use | Tag care-related screens; couple with weekly PHQ-4d |
| Social media/text only (Reddit, notes) [,] | Linguistic markers of stress, topics | Physical burden, sleep, adherence | Topic drift, platform bias, performative posts [] | Wearables + care task telemetry [] | Validates language with lived routines | Limit domain shift; time-align posts to passives |
| Chatbot-only (dialogue transcripts) [,] | Perceived support, coping language | Real adherence, off-chat distress | Hawthorne effect; “green-dot” compliance [] | Background passives, periodic scales [] | Detects off-session deterioration | Log safety phrases; escalate if passives worsen |
| Rule-based personalization (fixed heuristics) [,,,,,,] | Interpretability, safety constraints | Individual adaptation, nonstationarity | Stagnation; alert fatigue [,] | Bandits/RLe-lite + preference signals [] | Learns timing/content per user | Start conservative; log policy updates |
| Classical machine learning on small samples [-,,,,-] | Stability on tabular data | Complex temporal dynamics | Underfit, brittle thresholds [] | Sequence models + data augmentation [] | Captures trajectories; reduces variance | Time-aware CVf; report calibration |
| Deep learning only (black-box) [,,,-] | Nonlinear patterns, multimodal fusion | Clinician trust, auditability | Opaque errors; spurious cues [] | Post-hoc XAIg + policy logging + calibration [] | Explanations, traceability, reliable probabilities | Tiered transparency for high-stakes prompts |
| EHRh/clinic data only (retrospective) [] | Diagnoses, meds, visits | Home context, daily stressors | Label noise; delayed signals [,] | Home passives + care tasks + short EMAs [,] | Fills between visits; real-time risk signals | Federate; de-identify; align timelines |
| Scheduled questionnaires only (weekly/biweekly) [,] | Comparable scales over time | In-between episodes, burden spikes | Missed surveys; ceiling effects [] | Event-triggered micro-prompts + passives [] | JITi capture during anomalies | Use anomaly detectors to gate EMAs |
aHR: heart rate.
bHRV: heart rate variability.
cEMA: ecological momentary assessment.
dPHQ-4: 4-item Patient Health Questionnaire.
eRL: reinforcement learning.
fCV: cross-validation.
gXAI: explainable AI.
hEHR: electronic health record.
iJIT: just-in-time.
RQ1: Data Sources, Collection Methods, and Data Processing Techniques
Study Designs and Data Sources
The included studies reflect a diverse range of data sources, with the majority drawn from clinical trials, which accounted for 9 out of 28 (32.1%) studies. Mobile apps and smartphone-based systems represented 5 out of 28 (17.9%) studies, including platforms such as Meru Health, Moodie, MUBS, and Foundations, which captured self-reported mood tracking and passive sensor data. Open-source and social media datasets, including data from platforms like Reddit, Kaggle, and GitHub, appeared in 2 out of 28 (7.1%) studies. Wearable devices, such as smartwatches integrated with apps (eg, BrainE), were used in 1 (3.6%) study, as were AI-based mental health chatbots (eg, SAAC), structured surveys and interviews, and outpatient clinical data—each contributing 1 (3.6%) study. The remaining 8 out of 28 (28.6%) studies were classified as other or unspecified, comprising multimodal datasets, broadly described data sources, or those with insufficient detail to assign a specific category.
Collection Methods and Feature Extraction
The reviewed studies used a variety of data collection methods, including clinical trials in 3 (10.7%) studies [,,], mobile apps and smartphone-based systems in 2 (7.1%) studies [,], open-source or social media data in 2 (7.1%) studies [,], outpatient clinical data in 1 (3.6%) study [], surveys and interviews in 6 (21.4%) studies [,,,,,], feature extraction commonly included demographics and psychosocial traits in 9 (32.1%) studies [,,,-,,], sentiment and NLP-based features in 3 (10.7%) studies [,,], neurological and physiological metrics in 4 (14.2%) studies [,,,], while only 2 (7.1%) studies [,] focused on model accuracy and prediction.
Preprocessing and Integration Techniques
The reviewed studies used a wide range of preprocessing and integration methods. About 42.9% (12 out of 28) of the studies [,,,,-,,-] used at least one explicit preprocessing step. These steps included cleaning the data and dealing with missing values [,,], normalization, standardization, or encoding techniques [,,,,], and feature engineering, which included dichotomizing variables or extracting logs [,]. There were also mentions of advanced pipelines to keep data from leaking during cross-validation [,]. About 28.6% (8 out of 28) of studies [,,,,,,,] talked about combining multiple datasets or making different metrics work together. This included combining data from different demographics [,], wearable and app-based logs [,], and multimodal clinical sources [,,,]. The other studies (20/28, 71.4%) [,,,-,,,,,-] either used datasets from only one source or did not explain how they combined the data. These results show that there are big differences in how the literature reports on preprocessing transparency and integration strategies.
RQ2: Personalization Strategies and Real-Time AI-Driven Monitoring
Real-Time Monitoring Scope
The analysis of real-time monitoring in AI-driven mental health interventions was one of the primary goals of this review. In this review, 3 (10.7%) studies [,,] used real-time digital monitoring, such as through wearables or mobile apps; 2 (7.1%) studies [,] used scheduled mood or symptom questionnaires. Two (7.1%) studies used self-reported or passive sensor monitoring [,]. Two (7.1%) studies [,] used hybrid or unconventional methods, and 1 (3.6%) study reported periodic clinical follow-ups []. However, one significant gap in continuous tracking was highlighted by the fact that 18 out of 28 (64.3%) studies [-,-,,,,,-] either did not have any monitoring component or did not clearly report it. These findings indicate that real-time monitoring is underused, although early implementations indicate promising directions for future adaptive mental health systems.
Personalization and Intervention Strategies
Mood-based personalization was one of the most used defined methods; it was used in 4 studies [,,,]. Clinical stratification and risk profiling were used in 3 studies [,,]. However, 2 studies [,] used AI-powered therapeutic interfaces, while 2 studies [,] were on behavior-based personalization. In addition, 2 studies [,] implemented treatment response prediction. Moreover, general personalization frameworks without dynamic adaptation were present in 2 studies [,]. These patterns indicate growing interest but highlight the need for more consistent and transparent personalization models.
AI Methodology and Implementation Challenges
The studies we reviewed used a number of different AI technologies, such as traditional machine learning approaches [], including Random Forest and its variants like Boruta and Binary Mixed Model [,,,,,-]. Moreover, we found ensemble approaches combining models like Naïve Bayes, decision trees, support vector machine, logistic regression, linear models, and voting regressors [,,]. Furthermore, we saw several natural language processing techniques, including deep learning models such as recurrent neural networks, long short-term memory, gated recurrent unit, convolutional neural network, Efficiently Learning an Encoder that Classifies Token Replacements Accurately, and Bidirectional Encoder Representations from Transformers which were used for personalized recommendations [,,,-].
Several studies reported key implementation challenges. These included issues with interpretability in deep learning models [], data variability across sessions and algorithms []. Other issues were also found, such as chatbot systems' limited scope and privacy concerns [,]. Moreover, data imbalances related to antidepressant response also affected model performance [].
Critical Influencing Factors
The reviewed studies identified a wide range of influencing factors that affect the efficacy of AI-driven personalized mental health interventions. Among the eligible studies, 3 studies [,,] expressed concerns regarding treatment duration and adherence, and 3 others [,,] reported technical problems, such as small datasets or feature limitations. Besides, 2 studies [,] showed concerns about privacy and autonomy (such as hesitancy to share data or fear of overprediction) while dissatisfaction with excessive personalization and monitoring emerged in the studies by Chen et al [] and Alslaity et al []. Moreover, in 4 studies [,,,], demographic and psychosocial factors were highlighted. Other distinct factors included generalizability [], mood tracking [], sensor reliability [], and personalized risk factors []. These results emphasize the need for future AI systems for personalized mental health to have a multifaceted design and more transparent reporting.
Target Groups and Conditions
Although some studies focused on clinically diagnosed groups [-] and young adults or adolescents [,], most studies [,,,-,,,,] focused on adults between the ages of 18 and 65 years. Regional demographics were referenced by Kim et al [] and Jensen et al [], and a gender-specific focus surfaced in the study by Meinlschmidt et al []. The need for better reporting and broader inclusion in future research is highlighted by the fact that 8 studies [,-,,] lacked accurate population data.
RQ3: Explainability, Clinical Integration, and AI Decision Impact
Explainability Techniques
The reviewed studies used different approaches to explain AI. The majority of studies (21.4%) used tree-based and classical models [,,,,,], whereas 8 (28.6%) studies [-,,,,,] did not report any explanation approach. Interpretability of deep learning was mentioned in 2 (7.1%) studies [,], along with NLP-based tools [,], feature selection techniques [,], model evaluation visualizations [,], and feature importance tools [,]. Three (10.7%) studies used hybrid or other methods [,,], and 1 (3.6%) study reported explainability based on naive Bayes []. These findings highlight the need for more consistent and transparent AI systems in mental health by exposing unequal adoption of explainability practices.
Design and Implementation Challenges
The reviewed studies identified common deployment-related and technical issues in the design of AI systems for mental health. A notable issue was the lack of reporting in 14 (50.0%) studies [-,,,,-]. However, concerns regarding the integration of devices and sensors were shown in 2 (7.1%) studies, and 4 (14.3%) studies brought up the subject of data variability and quality [,,,]. Two (7.1%) studies reported privacy and scope limitations, particularly in chatbot systems [,]. Other challenges included misclassification and generalization gaps [], model overfitting [], user engagement issues [], and long-term trust erosion []. These findings show that more open and consistent reporting is needed to ensure the clinical adaptability of AI systems.
Trade-Offs Between Performance and Interpretability
The reviewed studies were markedly inconsistent with respect to the trade-off between explainability and AI model performance. Interpretability issues, particularly in deep learning models, were discovered in 4 (14.3%) studies [,,,]. Three (10.7%) studies raised broad concerns about finding a balance between clinical usability and accuracy [,,]. Two (7.1%) studies reported high-performing but opaque models [,], and 2 (7.1%) more emphasized the conflict between transparency and personalization [,]. Interestingly, no discussion of this topic was reported in 17 (60.7%) studies [-,,-]. These discrepancies highlight how explainable AI is necessary to improve clinical integration and trust.
Clinician-Centered Usability
Four studies [,,,] raised concerns regarding clinician acceptance and applicability, suggesting hesitancy to rely on AI systems without conclusive validation or clinical practice alignment. In 2 studies [,] that documented device dependency and relevance, wearable-specific data raised questions about clinical reliability. Lack of workflow integration [-], privacy and sensitivity restrictions [,], and restricted generalizability [] were further difficulties. Only one study [] specifically addressed explainability issues, and another [] raised doubts about the effectiveness of the intervention. A significant gap exists in addressing real-world integration, as evidenced by the 13 (46.4%) studies [,,,-,,,-] that did not identify any clinician-related concerns. These results highlight how crucial it is to build AI systems that promote clinician trust, usability, and practicality.
Impact on Decision-Making
In our review, studies showed different but promising effects on making clinical decisions. About 32.1% (9 out of 28) [,,,,-,,] agreed that depression should be diagnosed and its severity predicted. About 21.4% (6 out of 28) [,,-,] were focused on giving personalized treatment recommendations. At the same time, 10.7% (3/28) studies [,,] focused on chatbot-based or digital therapist roles, and 7.1% (2/28) studies [,] reported that patient-centered outcomes like planning daily activities and reducing symptoms had improved more effectively. The findings indicate that AI is becoming increasingly crucial in improving mental health care, but they also show that more real-world testing is needed.
Transitioning From Systematic Findings to Framework Design
Building upon the findings from this systematic review, we identified persistent challenges within the realm of AI-driven mental health interventions. Specifically, the lack of adequate personalization, minimal explainability, and insufficient integration into established clinical workflows. These deficiencies are particularly significant in high-burden health care domains, such as AD/ADRD, where caregivers contend with both emotional distress and intricate informational demands. The insights we gathered from analyzing different studies demonstrated the importance of developing a more adaptive, transparent, and contextually aware AI framework. In the subsequent sections, we have used our findings to propose a dynamic personalization and explainable decision-making enabled human-AI collaborative system for caregivers of individuals with ADRD.
Discussion
Principal Findings
In this review, we analyzed 28 papers on personalized AI-powered mental health treatment strategies by examining the types of data used to detect early warning signs of mental disorders. We also explored how AI-enabled systems deliver real-time monitoring and dynamically personalize treatment plans based on individual user behavior and engagement patterns. Moreover, we identified the challenges associated with designing and validating explainable AI models. We also discussed the limitations of the reviewed studies following our RQs, to identify key gaps and opportunities for future work, as outlined below.
RQ1: Limitations in Data Sources and Processing Practices
When designing personalized mental health intervention tools, there is no “one-size-fits-all” solution, particularly when it comes to selecting and tailoring input data sources []. In our review, it was visible that different studies used different approaches. In our review, though we found a large number of studies used self-reported questionnaires to get information about mood and symptoms [,,,,,], there are some major issues associated with this type of input. For instance, Stone et al [] highlighted some of the major issues in explaining EMA in behavioral medicine. According to the author, a major issue with traditional self-reported questionnaires is recall bias, which means that people might forget, underreport, or misrepresent how they felt or what symptoms they had in the past. Another issue is social desirability bias, which happens when people change their answers to make them seem more acceptable or favorable []. Furthermore, some studies used NLP-derived textual features [,,] to include detailed insights. However, these studies might face issues including contextual ambiguity [], domain adaptation [], and ethical concerns related to privacy and explainability [,]. The key limitations of the key data types are shown together in .

In addition to these choices for collecting data, our review also found that reporting practices were not very clear or consistent. There was a significant gap with enough details about the steps for preprocessing and integrating data. For example, only a few of the 6 (21.4%) studies that used surveys and interviews [,,,,,] clearly explained how these were processed or combined with other types of data. These gaps make it hard to see important methodological choices about data cleaning, normalization, and integration that can have a significant impact on AI performance [,]. Last, we found that methodological integration and performance evaluation were underexplored. For instance, only 2 studies [,] looked at how to combine different types of data, like clinical records with smartphone-based sensing or wearable-derived physiological signals. In the same way, only 2 (7.1%) studies [,] looked at model performance indicators like accuracy and prediction.
RQ2: Limitations in Personalization Strategies and Real-Time Monitoring
In the last several years, mental health systems have been increasingly moving toward personalization based on the idea that no 2 people deal with mental health issues in the same manner []. In personalization, real-time monitoring is one of the crucial variables. Traditional static AI mental health care approaches rely on retrospective or infrequent assessments that fail to capture dynamic, context-dependent symptom changes. Real-time monitoring solves this problem by gathering continuous, real-world data on mood and behavior []. Therefore, real-time monitoring tools enable digital mental health technologies that are aware of the environment to provide more flexible and effective treatment []. However, in our review, we found that even though real-time monitoring tools are essential, they were not widely used or used consistently in our studies. Instead, many studies used static or scheduled assessments that are incapable of detecting quick mood changes or environmental triggers. For example, studies [,] used retrospective mood questionnaires that were given weekly or at clinic visits, which are unable to inform timely, personalized interventions.
In our review, different personalization methods were found in mental health AI interventions, such as mood-based personalization [,,,], clinical stratification or risk profiling [,,,], and general personalization frameworks without dynamic adaptation [,]. However, all these personalization methods have some drawbacks. For instance, mood-based personalization often uses self-reported ratings from a single source that fails to consider evolving, contextual, or physiological factors that are important for adaptive mental health support [,]. In addition, clinical stratification and risk profiling often use static baseline assessments that disregard considering behavior or physiology, which makes adaptive prediction less accurate [,]. General personalization frameworks, on the other hand, use fixed rules that cannot change when users do, which makes them less relevant and engaging []. These limitations indicate significant gaps where multimodal data integration might be beneficial. Multimodal data integration can provide a comprehensive view of mental health states by combining different types of data, like physiological signals from wearables, self-reported mood ratings, and text written by users.
RQ3: Limitations in Explainability, Generalizability, and Clinical Integration
Explainability is one of the main ingredients of personalization because it helps users understand and trust AI-driven decisions. For instance, in the study by Kaur et al [], participants used a bird identification system with 4 different explainable AI methods, and reported that explanations helped them calibrate their trust in the AI’s output and improve their own identification skills. This finding illustrates how meaningful explanations can increase user confidence and support effective use of AI systems, a principle that is equally important in sensitive contexts like mental health care. In our review, we found that while many studies used tree-based and classical models for inherent interpretability [,,,,,], others used deep learning interpretability [,], NLP-based tools [,], feature selection [,], model evaluation visualizations [,], feature importance tools [,], hybrid methods [,,], and naive Bayes approaches []. However, several studies did not report any explainability approach at all [-,,,,,]. Therefore, this highlights the ongoing need for more consistent and thoughtful use of explainability techniques in mental health AI systems.
Our review revealed that few studies adequately addressed the issue of generalizability. The study by Webb et al [] specifically highlighted that models trained on limited or homogeneous data may demonstrate suboptimal performance across varied demographic, cultural, or clinical contexts, thereby emphasizing the risk of inequitable outcomes in real-world mental health care. Besides, since this limitation is rooted in the diversity of people whose data is used, inputs such as wearables, mood-based self-reports, and text data alone do not inherently guarantee broader generalizability []. At the same time, another critical challenge in developing AI for mental health is balancing predictive performance with interpretability. Several studies in our analysis chose high-performing but black-box models [,] without sufficiently addressing how their lack of transparency may undermine clinician confidence or safe implementation. Another issue that is quite common in AI interventions related to mental health is the trade-off between model accuracy and practical usability. For instance, a limited number of studies [,,] clearly addressed the trade-off between model accuracy and practical usability, underscoring a deficiency in design thinking crucial for real-world implementation. However, users often choose systems that are somewhat less accurate but more understandable, since they enhance comprehension, trust calibration, and informed decision-making []. Therefore, carefully addressing this trade-off is essential to guarantee that AI systems provide clinically significant help while maintaining transparency and acceptability for both practitioners and patients.
Proposed Conceptual Framework
Overview
In this study, we are proposing a conceptual framework for personalized mental health care specifically for caregivers of patients with AD/ADRD. This framework is based on the findings of our systematic review of AI-powered personalized mental health treatment or care interventions.
In our conceptual framework, we propose integrating wearable-derived physiological signals, mood-based self-reports, and free-text entries describing users’ feelings to capture multidimensional, context-rich data that supports personalized mental health interventions. This approach emerges as complementary based on the evidence in . By following EMA [], the framework is designed to take the user input multiple times daily. These multimodal data, followed by EMA, will solve the issues that we found in our review, such as recall bias and social desirability bias. However, this proposed framework has primarily 3 main features, such as a daily routine scheduler, a digital reward system, and an educational repository, as shown in . This framework represents a significant advancement over prior approaches, which have primarily concentrated on developing conceptual frameworks or taxonomies within the field. The whole overview of our framework is shown in .


Daily Routine Scheduler
A large portion of the routine of caregivers of individuals with AD/ADRD consists of the patient’s daily activities. Therefore, if caregivers are given a daily routine scheduler that combines the necessary needs and schedules of patients with AD/ADRD, it will be more beneficial for them. Keeping that in mind, the daily routine scheduler feature is composed of some subfeatures. First, there will be time‑stamped checklists for activities of daily living of patients with AD/ADRD, such as medications, meals, hygiene, and exercises. This will reduce the cognitive load of caregivers and ensure the attendance of critical things such as medication dosage.
Second, by enhancing the intrinsic motivation of self-determination theory [], the system will have premade templates according to the caregiver’s input. According to Jin et al [], allowing users to modify certain things in the system and receiving encouragement while doing so (Autonomy) is an important factor. Therefore, in our framework, a premade template will be given to the caregivers, and they will have the autonomy of alterations according to their preferences. When the caregivers make any changes to the templates, the system will learn from the changes, and it will provide a better template next time by using those learnings. Additionally, the system will provide suggestions that adjust task durations and prompts over time according to mild, moderate, and severe Alzheimer. This will match the changing abilities without the caregiver manually building schedules.
Third, to support the caregiver’s wellbeing, there will be self-care routine which will have scheduled self‑care breaks, hydration, short exercises, mood check‑in prompts that feed a private mental‑health log, which combinedly promotes caregiver health and flags early burnout. Through the mood checking prompts, the app will suggest some low-burden, context-aware prompts, such as a one‑tap breathing exercise, a 5‑minute stretch video, or a brief “Pause and hydrate” reminder right on the home screen, which shows the relatedness []. Therefore, together, these features turn the scheduler into a smart, caregiver‑centered system that balances patient care, caregiver’s health, and easy coordination with the wider support network.
Digital Incentive and Reward Framework
According to Jin et al [], the user needs a sense of achievement to be motivated using a system, which is called competence. To ensure competence, the framework provides a feature of getting digital rewards, which will turn routine duties and self-care into incremental accomplishments, countering burnout by providing immediate, tangible affirmation. Whenever a caregiver completes a scheduled task, such as achieving milestones by logging mood checking, keeping a certain streak, 30-minute walks for 7 days, etc, a digital reward will be provided, which they might show as their progress to social media to encourage others.
Digital Educational Content Repository Framework
Despite having formal caregivers, family members are often involved in the care of patients with AD/ADRD as informal caregivers. Most of them do not have the proper training and knowledge to handle patients with proper care. Besides, their busy schedules do not allow them to get proper training or even to acquire knowledge from different sources. In that case, our app will help them immensely. To help them get the knowledge of patients with AD/ADRD–related articles or seminars, the app will provide them with bite-sized knowledge, which will be easy to adopt and time-efficient. If they want to know more about a specific article, they can go to the link by simply clicking on the bite-sized learning and get the full article. Moreover, our system will provide all related information, such as seminars, articles, and other educational sources, together for a better learning experience, which will decrease the workload of searching for necessary information in distinct locations.
Future Directions
From the findings of the review, we found that the existing personalized mental health–related AI approaches have significant shortcomings in terms of explainability, transparency, and usability testing. In that case, our proposed framework directly addresses these deficiencies by furnishing a clear, user‑specific rationale for every recommendation—for example, a prompt suggesting a 20‑minute walk is justified by noting that the caregiver previously reported this activity as effective during periods of elevated stress. By consistently articulating the reasoning behind each action, the system enhances transparency, cultivates user trust, and supports sustained engagement.
The development of this conceptual framework presents profound opportunities for future work. Further research might incorporate the technology acceptance model to evaluate the adoption and usability of the proposed system. This will help to understand how users will accept and use this technology, focusing on 2 primary constructs such as perceived usefulness and perceived ease of use. Besides, according to Kushniruk and Patel [], usability testing enhances generalizability by uncovering obstacles and design deficiencies across many user groups and circumstances, which will overcome the limitations found in the review. Therefore, deploying the framework in real-world environments will allow us to assess its scalability, integration with institutional routines, and effectiveness in supporting professional caregivers alongside family members.
Limitations
This study has several notable limitations. First, none of the included studies were about patients with AD/ADRD or targeted caregivers. The studies were more focused on AI-driven personalized mental health care among the general population. However, this also serves as one of the strengths of this study. By analyzing the global advancement in personalized AI mental health interventions, we will be able to apply the outcome of the extracted foundational principles, such as mood-based personalization and explainability, from the review. This allowed us to design a forward-looking framework that can be adapted and validated in future contexts specific to caregivers of individuals with AD/ADRD.
Second, the proposed framework has yet to be evaluated through broader lenses with external experts. Third, there is a profound need for real-world usability studies to assess how caregivers interact with the system, perceive its usefulness, and engage with its adaptive and emotional support features. This has been reserved as one of the future works that will help the system to be fine-tuned.
Fourth, despite the review focused on studies that implemented AI-based personalized mental health interventions, we also included several papers in which the authors proposed that their models could potentially support personalization in future applications. This also worked as a strength of this study, as this enabled us to capture emerging perspectives and early-stage innovations, which kept us updated about the future landscape of AI-driven mental health care.
Fifth, the framework is currently suitable for home-based settings. However, we aim to facilitate the framework in a facility-based care setting and incorporate it with clinical experts. Finally, the review was done with studies that were published between January 2020 and February 2025; yet, new studies are continuing to be published. Therefore, an updated review of AI-driven mental health care interventions may be warranted in the near future.
Despite having these limitations, our proposed framework, especially on personalized mental health research on caregivers of patients with AD/ADRD, serves as a technically robust and future-proof design. However, future work is needed to check user acceptance and tailor these findings through caregiver-specific research.
Acknowledgments
This work was supported by the mHealth Lab at Kennesaw State University.
Authors' Contributions
SUS conceptualized the overall design of the work and wrote this manuscript. SUS was also involved in designing and conducting the search. SUS, CRR, and IS were involved in screening results and extracting data. All authors contributed to the final version of the manuscript and approved the submission.
Conflicts of Interest
None declared.
PRISMA checklist.
PDF File (Adobe PDF File), 64 KBSearch query.
PDF File (Adobe PDF File), 93 KBDetailed data sources, feature extraction, and processing methods.
PDF File (Adobe PDF File), 381 KBDetailed personalization strategies and real-time AI-driven monitoring.
PDF File (Adobe PDF File), 369 KBDetailed explainability, clinical Integration, and impact of artificial intelligence–mediated decisions.
DOCX File , 30 KBReferences
- Mather M, Jacobsen L, Pollard K. Aging baby boomers to face caregiving, obesity, inequality challenges. PRB. 2016. URL: https://www.prb.org/resources/aging-baby-boomers-to-face-caregiving-obesity-inequality-challenges/ [accessed 2025-01-13]
- 2023 alzheimer's disease facts and figures. Alzheimer's Association. URL: https://www.alz.org/alzheimers-dementia/facts-figures [accessed 2025-02-20]
- Sallim AB, Sayampanathan AA, Cuttilan A, Ho R. Prevalence of mental health disorders among caregivers of patients with alzheimer disease. J Am Med Dir Assoc. 2015;16(12):1034-1041. [CrossRef] [Medline]
- Schulz R, Sherwood P. Physical and mental health effects of family caregiving. Am J Nurs. 2008;108(9 Suppl):23-7; quiz 27. [FREE Full text] [CrossRef] [Medline]
- Educational programs and dementia care resources. Alzheimer's Association. Chicago, IL.; 2019. URL: https://www.alz.org/help-support/resources/care-education-resources [accessed 2026-01-13]
- Alzheimer's Universe. URL: https://www.alzu.org/ [accessed 2025-01-13]
- Find a therapist. Therapy tribe. URL: https://www.therapytribe.com/ [accessed 2026-01-13]
- Family Caregiver Alliance. San Francisco, CA. Caregiver.org; 2020. URL: https://www.caregiver.org/ [accessed 2025-05-07]
- Dementia talking point. Alzheimer's Society Dementia Support Forum. 2019. URL: https://forum.alzheimers.org.uk/ [accessed 2025-05-07]
- Dementia CareAssist app. National Institute on Aging. URL: https://www.nia.nih.gov/research/alzheimers-dementia-outreach-recruitment-engagement-resources/dementia-careassist-app [accessed 2024-09-03]
- Samsung's new family care service helps caregivers monitor loved ones. The Verge. 2024. URL: https://www.theverge.com/2024/8/14/24219489/samsung-smartthings-family-care-caregiver-support-launch-price-release-date [accessed 2025-01-13]
- Fan Q, Hoang M-N, DuBose L, Ory MG, Vennatt J, Salha D, et al. The Olera.care digital caregiving assistance platform for dementia caregivers: preliminary evaluation study. JMIR Aging. Apr 17, 2024;7:e55132. [FREE Full text] [CrossRef] [Medline]
- Amershi S, Weld D, Vorvoreanu M, Fourney A, Nushi B, Collisson P. Guidelines for human-AI interaction. 2019. Presented at: CHI '19: CHI Conference on Human Factors in Computing Systems; 2019 May 4 - 9:1-13; Glasgow Scotland Uk. URL: https://doi.org/10.1145/3290605.3300233 [CrossRef]
- Jung J, Kang S, Choi J, El-Kareh R, Lee H, Kim H. Evaluating the impact of explainable AI on clinicians' decision-making: a study on ICU length of stay prediction. Int J Med Inform. Sep 2025;201:105943. [FREE Full text] [CrossRef] [Medline]
- Belle S, Burgio L, Burns R, Coon D, Czaja SJ, Gallagher-Thompson D, et al. Resources for Enhancing Alzheimer's Caregiver Health (REACH) II Investigators. Enhancing the quality of life of dementia caregivers from different ethnic or racial groups: a randomized, controlled trial. Ann Intern Med. 2006;145(10):727-738. [FREE Full text] [CrossRef] [Medline]
- Mittelman M, Ferris SH, Shulman E, Steinberg G, Levin B. A family intervention to delay nursing home placement of patients with alzheimer disease. a randomized controlled trial. JAMA. 1996;276(21):1725-1731. [Medline]
- Schulz R, Eden J. Family Caregiving Roles and Impacts. Washington, DC. National Academies Press (US); 2016.
- Pearlin L, Mullan JT, Semple SJ, Skaff MM. Caregiving and the stress process: an overview of concepts and their measures. Gerontologist. 1990;30(5):583-594. [CrossRef] [Medline]
- Zarit S, Reever KE, Bach-Peterson J. Relatives of the impaired elderly: correlates of feelings of burden. Gerontologist. 1980;20(6):649-655. [CrossRef] [Medline]
- Nahum-Shani I, Smith SN, Spring BJ, Collins LM, Witkiewitz K, Tewari A, et al. Just-in-time adaptive interventions (JITAIs) in mobile health: key components and design principles for ongoing health behavior support. Ann Behav Med. 2018;52(6):446-462. [FREE Full text] [CrossRef] [Medline]
- Klasnja P, Hekler EB, Shiffman S, Boruvka A, Almirall D, Tewari A, et al. Microrandomized trials: an experimental design for developing just-in-time adaptive interventions. Health Psychol. 2015;34S:1220-1228. [FREE Full text] [CrossRef] [Medline]
- Kim K, Ryu JI, Lee BJ, Na E, Xiang Y, Kanba S, et al. A machine-learning-algorithm-based prediction model for psychotic symptoms in patients with depressive disorder. J Pers Med. 2022;12(8):1218. [FREE Full text] [CrossRef] [Medline]
- Rubel JA, Zilcha-Mano S, Giesemann J, Prinz J, Lutz W. Predicting personalized process-outcome associations in psychotherapy using machine learning approaches-a demonstration. Psychother Res. 2020;30(3):300-309. [CrossRef] [Medline]
- Monaco F, Vignapiano A, Piacente M, Farina F, Pagano C, Marenna A, et al. Innova4Health: an integrated approach for prevention of recurrence and personalized treatment of major depressive disorder. Front Artif Intell. 2024;7:1366055. [FREE Full text] [CrossRef] [Medline]
- Jensen KHR, Dam VH, Ganz M, Fisher PM, Ip C, Sankar A, et al. Deep phenotyping towards precision psychiatry of first-episode depression - the brain drugs-depression cohort. BMC Psychiatry. 2023;23(1):151. [FREE Full text] [CrossRef] [Medline]
- Amirhosseini MH, Ayodele AL, Karami A. Prediction of depression severity and personalised risk factors using machine learning on multimodal data. IEEE; 2024. Presented at: IEEE 12th International Conference on Intelligent Systems (IS); 2024 August 29-31:1-7; Varna, Bulgaria. [CrossRef]
- Shani R, Tal S, Derakshan N, Cohen N, Enock PM, McNally RJ, et al. Personalized cognitive training: protocol for individual-level meta-analysis implementing machine learning methods. J Psychiatr Res. 2021;138:342-348. [FREE Full text] [CrossRef] [Medline]
- Vijayaraghavan A, Radhakrishnan J, Karthik RA, Ceralaathan G, Sailesh S. AI powered chatbot for mental health treatment. 2024. Presented at: 2024 First International Conference on Technological Innovations and Advance Computing (TIACOMP); 2024 June 29-30; Bali, Indonesia. [CrossRef]
- Pandey S, Sharma S, Wazir S. Mental healthcare chatbot based on natural language processing and deep learning approaches: Ted the therapist. Int J Inf Technol. 2022;14(7):3757-3766. [CrossRef]
- Ravichand M, Singh J, Shelke N, Ramesh G, Srivastava S, Muniyandy E. Evaluating the efficacy of deep learning models in personalizing treatment for anxiety disorders. 2024. Presented at: 2024 4th International Conference on Intelligent Technologies (CONIT); 2024 June 21-23:1-6; Hubballi, Karnataka, India. [CrossRef]
- Priyanka S, Revathi K, Priyadharshini S, Niranjan NN, Visalachi N, Sumathi P. AI-driven approaches to enhancing mental wellbeing and stress relief. 2025. Presented at: 2025 International Conference on Multi-Agent Systems for Collaborative Intelligence (ICMSCI); 2025 January 22:925-931; Erode, Tamil Nadu, India. [CrossRef]
- Kavyashree N, Usha J. MediBot: healthcare assistant on mental health and well being. IEEE; 2023. Presented at: 7th International Conference on Computation System and Information Technology for Sustainable Solutions (CSITSS); 2023 November 02-04:1-7; Bangalore, India. [CrossRef]
- Filippis RD, Foysal AA. Advanced machine learning models for gender-specific antidepressant response prediction: overcoming data imbalance for precision psychiatry. OALib. 2025;12(02):1-13. [CrossRef]
- Hornstein S, Forman-Hoffman V, Nazander A, Ranta K, Hilbert K. Predicting therapy outcome in a digital mental health intervention for depression and anxiety: a machine learning approach. Digit Health. 2021;7:20552076211060659. [FREE Full text] [CrossRef] [Medline]
- Lipschitz JM, Lin S, Saghafian S, Pike CK, Burdick KE. Digital phenotyping in bipolar disorder: using longitudinal Fitbit data and personalized machine learning to predict mood symptomatology. Acta Psychiatr Scand. 2025;151(3):434-447. [CrossRef] [Medline]
- M. Eid M, Yundong W, Benneh Mensah G, Pudasaini P. Treating psychological depression utilising artificial intelligence: AI for precision medicine- focus on procedures. Mesopotam J Artif Intell Healthc. 2023;2023:76-81. [CrossRef]
- Shah RV, Grennan G, Zafar-Khan M, Alim F, Dey S, Ramanathan D, et al. Personalized machine learning of depressed mood using wearables. Transl Psychiatry. 2021;11(1):338. [FREE Full text] [CrossRef] [Medline]
- Benrimoh D, Kleinerman A, Furukawa TA, Iii CFR, Lenze EJ, Karp J, et al. Towards outcome-driven patient subgroups: a machine learning analysis across six depression treatment studies. Am J Geriatr Psychiatry. 2024;32(3):280-292. [CrossRef] [Medline]
- Meinlschmidt G, Tegethoff M, Belardi A, Stalujanis E, Oh M, Jung EK, et al. Personalized prediction of smartphone-based psychotherapeutic micro-intervention success using machine learning. J Affect Disord. 2020;264:430-437. [CrossRef] [Medline]
- Doe J, Smith A, Lee B. MUBS: a personalized recommender system for behavioral activation in mental health. 2020. Presented at: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems; April 25-30, 2020:1-14; Honolulu, United States. [CrossRef]
- Johnson K, Williams M, Zhao Y. Personalized recommendations in mental health apps: the impact of autonomy and data sharing. In.; 2021. Presented at: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems; May 8-13, 2021:1-8; Yokohama, Japan. [CrossRef]
- Chen L, Kim H, Patel R. Exploring the effects of AI-assisted emotional support processes in online mental health communities. 2022. Presented at: Proceedings of the CHI Conference on Human Factors in Computing Systems Extended Abstracts; April 29-May 5, 2022:1-7; New Orleans, LA, USA. [CrossRef]
- Alslaity A, Chan G, Orji R. Insights from longitudinal evaluation of Moodie mental health app. 2022. Presented at: CHI '22: CHI Conference on Human Factors in Computing Systems; 29 April- 5 May, 2022:1-8; New Orleans LA USA. [CrossRef]
- Sharma MK, Nachappa MN, Kumar R. Personalized treatment recommendations for mental health disorders using AI and big healthcare data. IEEE; 2023. Presented at: IEEE International Conference on ICT in Business Industry & Government (ICTBIG); 2023 December 08-09:1-6; Indore, India. [CrossRef]
- Webb CA, Hirshberg MJ, Davidson RJ, Goldberg SB. Personalized prediction of response to smartphone-delivered meditation training: randomized controlled trial. J Med Internet Res. 2022;24(11):e41566. [FREE Full text] [CrossRef] [Medline]
- van Bronswijk SC, DeRubeis RJ, Lemmens LHJM, Peeters FPML, Keefe JR, Cohen ZD, et al. Precision medicine for long-term depression outcomes using the personalized advantage index approach: cognitive therapy or interpersonal psychotherapy? Psychol Med. 2021;51(2):279-289. [FREE Full text] [CrossRef] [Medline]
- Ramzan HA, Abdulah F, Ahmad M, Ramzan S, Ashraf M. AI-driven personalization of e-therapy interventions for anxiety, stress, and depression. 2024. Presented at: 18th International Conference on Open Source Systems and Technologies (ICOSST); 2024 December 26-27:1-6; Lahore, Pakistan. [CrossRef]
- Schmidt F, Hammerfald K, Jahren H. Using machine learning to recommend personalized modular treatments for common mental health disorders. 2023. Presented at: IEEE International Conference on Digital Health (ICDH); IEEE:150-157; Chicago, IL, USA. [CrossRef]
- Pushpa G, Chaitra M, Kolur LP, Dhananjaya S, Kavyasri MN, Sunitha R, et al. An advanced AI framework for mental health diagnostics using bidirectional encoder representations from transformers with gated recurrent units and convolutional neural networks. ISI. 2025;30(1):213-220. [CrossRef]
- Stone AA, Shiffman S. Capturing momentary, self-report data: a proposal for reporting guidelines. Ann Behav Med. 2002;24(3):236-243. [CrossRef] [Medline]
- Jones A, Remmerswaal D, Verveer I, Robinson E, Franken IHA, Wen CKF, et al. Compliance with ecological momentary assessment protocols in substance users: a meta-analysis. Addiction. 2019;114(4):609-619. [FREE Full text] [CrossRef] [Medline]
- Cornet VP, Holden RJ. Systematic review of smartphone-based passive sensing for health and wellbeing. J Biomed Inform. 2018;77:120-132. [FREE Full text] [CrossRef] [Medline]
- Weed L, Lok R, Chawra D, Zeitzer J. The impact of missing data and imputation methods on the analysis of 24-hour activity patterns. Clocks Sleep. 2022;4(4):497-507. [FREE Full text] [CrossRef] [Medline]
- Slyepchenko A, Uher R, Ho K, Hassel S, Matthews C, Lukus PK, et al. A standardized workflow for long-term longitudinal actigraphy data processing using one year of continuous actigraphy from the CAN-BIND wellness monitoring study. Sci Rep. 2023;13(1):15300. [FREE Full text] [CrossRef] [Medline]
- Baumel A, Muench F, Edan S, Kane JM. Objective user engagement with mental health apps: systematic search and panel-based usage analysis. J Med Internet Res. 2019;21(9):e14567. [FREE Full text] [CrossRef] [Medline]
- Eysenbach G. The law of attrition. J Med Internet Res. 2005;7(1):e11. [FREE Full text] [CrossRef] [Medline]
- Kroenke K, Spitzer RL, Williams JBW, Löwe B. An ultra-brief screening scale for anxiety and depression: the PHQ-4. Psychosomatics. 2009;50(6):613-621. [CrossRef] [Medline]
- Chancellor S, De Choudhury M. Methods in predictive techniques for mental health status on social media: a critical review. npj Digit Med. 2020;3:43. [FREE Full text] [CrossRef] [Medline]
- Lee J, Browning E, Hokayem J, Albrechta H, Goodman GR, Venkatasubramanian K, et al. Smartphone and wearable device-based digital phenotyping to understand substance use and its syndemics. J Med Toxicol. 2024;20(2):205-214. [CrossRef] [Medline]
- McCambridge J, Witton J, Elbourne DR. Systematic review of the Hawthorne effect: new concepts are needed to study research participation effects. J Clin Epidemiol. 2014;67(3):267-277. [FREE Full text] [CrossRef] [Medline]
- Thayer JG, Franklin A, Miller JM, Grundmeier RW, Rogith D, Wright A. A scoping review of rule-based clinical decision support malfunctions. J Am Med Inform Assoc. 2024;31(10):2405-2413. [CrossRef] [Medline]
- Thayer J, Franklin A, Miller JM, Grundmeier RW, Rogith D, Wright A. A scoping review of rule-based clinical decision support malfunctions. J Am Med Inform Assoc. 2024;31(10):2405-2413. [CrossRef] [Medline]
- Liao P, Greenewald K, Klasnja P, Murphy S. Personalized heartsteps: a reinforcement learning algorithm for optimizing physical activity. Proc ACM Interact Mob Wearable Ubiquitous Technol. 2020;4(1):1. [FREE Full text] [CrossRef] [Medline]
- Riley R, Collins GS. Stability of clinical prediction models developed using statistical or machine learning methods. Biom J. 2023;65(8):e2200302. [FREE Full text] [CrossRef] [Medline]
- Ahmed D, Hassan MM, Mstafa RJ. A review on deep sequential models for forecasting time series data. Appl Comput Intell Soft Comput. 2022;2022:1-19. [FREE Full text] [CrossRef]
- Zech J, Badgeley MA, Liu M, Costa AB, Titano JJ, Oermann EK. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study. PLoS Med. 2018;15(11):e1002683. [FREE Full text] [CrossRef] [Medline]
- Ribeiro MT, Singh S, Guestrin C. "Why Should I Trust You?": explaining the predictions of any classifier. 2016. Presented at: KDD '16: The 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2016 August 13 - 17:1135-1144; San Francisco California USA. URL: https://doi.org/10.1145/2939672.2939778 [CrossRef]
- O'Malley KJ, Cook KF, Price MD, Wildes KR, Hurdle JF, Ashton CM. Measuring diagnoses: ICD code accuracy. Health Serv Res. 2005;40(5 Pt 2):1620-1639. [FREE Full text] [CrossRef] [Medline]
- Weiskopf N, Weng C. Methods and dimensions of electronic health record data quality assessment: enabling reuse for clinical research. J Am Med Inform Assoc. 2013;20(1):144-151. [FREE Full text] [CrossRef] [Medline]
- Bufano P, Laurino M, Said S, Tognetti A, Menicucci D. Digital phenotyping for monitoring mental disorders: systematic review. J Med Internet Res. 2023;25:e46778. [FREE Full text] [CrossRef] [Medline]
- Torous J, Kiang MV, Lorme J, Onnela J-P. New tools for new research in psychiatry: a scalable and customizable platform to empower data driven smartphone research. JMIR Ment Health. 2016;3(2):e16. [FREE Full text] [CrossRef] [Medline]
- Mohr D, Zhang M, Schueller S. Personal Sensing: understanding mental health using ubiquitous sensors and machine learning. Annu Rev Clin Psychol. 2017;13:23-47. [FREE Full text] [CrossRef] [Medline]
- Stone AA, Shiffman S, Schwartz JE, Broderick JE, Hufford MR. Patient compliance with paper and electronic diaries. Control Clin Trials. 2003;24(2):182-199. [CrossRef] [Medline]
- Van DMT. Faking it: social desirability response bias in self-report research. Aust J Adv Nurs. 2008:40-48. [FREE Full text]
- Singh G. Sentiment analysis of code-mixed social media text (Hinglish). arXiv:2102.12149. 2021. [FREE Full text]
- Gururangan S, Marasović A, Swayamdipta S, Lo K, Beltagy I, Downey D, et al. Don't stop pretraining: adapt language models to domains and tasks. 2026. Presented at: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2026 January 12:8342-8360; Online. URL: https://doi.org/10.48550/arXiv.2004.10964 [CrossRef]
- Benton A, Coppersmith G, Dredze M. Ethical research protocols for social media health research. Assoc Comput Linguist. 2017:94-102. [FREE Full text] [CrossRef]
- Chancellor S, Pater JA, Clear T, Gilbert E, De CM. Multimodal classification of moderated online pro-eating disorder content. 2017. Presented at: CHI '17: CHI Conference on Human Factors in Computing Systems; 2017 May 6 - 11:3213-32226; Denver Colorado USA. URL: https://doi.org/10.1145/3025453.3025985 [CrossRef]
- Liu X, Rivera S, Moher D, Calvert M, Denniston AK, SPIRIT-AICONSORT-AI Working Group. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. BMJ. 2020;370:m3164. [FREE Full text] [CrossRef] [Medline]
- Roberts M, Driggs D, Thorpe M, Gilbey J, Yeung M, Ursprung S, et al. Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans. Nat Mach Intell. 2021;3(3):199-217. [FREE Full text] [CrossRef]
- Shiffman S, Stone A, Hufford M. Ecological momentary assessment. Annu Rev Clin Psychol. 2008;4:1-32. [CrossRef] [Medline]
- Bardram JE, Frost M, Szántó K, Faurholt-Jepsen M, Vinberg M, Kessing LV. Designing mobile health technology for bipolar disorder. 2013. Presented at: CHI '13: CHI Conference on Human Factors in Computing Systems; 2013 27 April - 2 May:2627-2637; Paris France. URL: https://doi.org/10.1145/2470654.2481364 [CrossRef]
- Hekler EB, Klasnja P, Chevance G, Golaszewski N, Lewis D, Sim I. Why we need a small data paradigm. BMC Med. 2019;17(1):133. [FREE Full text] [CrossRef] [Medline]
- Mohr DC, Weingardt KR, Reddy M, Schueller SM. Three problems with current digital mental health research . . . and three things we can do about them. Psychiatr Serv. 2017;68(5):427-429. [FREE Full text] [CrossRef] [Medline]
- Torous J, Wisniewski H, Liu G, Keshavan M. Mental health mobile phone app usage, concerns, and benefits among psychiatric outpatients: comparative survey study. JMIR Ment Health. 2018;5(4):e11715. [FREE Full text] [CrossRef] [Medline]
- Fitzpatrick K, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (woebot): a randomized controlled trial. JMIR Ment Health. 2017;4(2):e19. [FREE Full text] [CrossRef] [Medline]
- Kaur H, Nori H, Jenkins S, Caruana R, Wallach H, Wortman VJ. Interpreting interpretability: understanding data scientists' use of interpretability tools for machine learning. 2020. Presented at: CHI '20: CHI Conference on Human Factors in Computing Systems; 2020 April 25 - 30:1-14; Honolulu HI USA. URL: https://doi.org/10.1145/3313831.3376219 [CrossRef]
- Cornet VP, Holden RJ. Systematic review of smartphone-based passive sensing for health and wellbeing. J Biomed Inform. 2018;77:120-132. [FREE Full text] [CrossRef] [Medline]
- Jin S, Hong H, Park J, Lee U. "I Don't Know Why I Should Use This App": holistic analysis on user engagement challenges in mobile mental health. 2025. Presented at: CHI '25: Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems; 2025 26 April- 1 May:1-23; Yokohama Japan. URL: https://doi.org/10.1145/3706598.3713732 [CrossRef]
- Kushniruk AW, Patel VL. Cognitive and usability engineering methods for the evaluation of clinical information systems. J Biomed Inform. 2004;37(1):56-76. [FREE Full text] [CrossRef] [Medline]
Abbreviations
| AD/ADRD: Alzheimer disease and related dementias |
| AI: artificial intelligence |
| EMA: ecological momentary assessment |
| NLP: natural language processing |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
| RQ: research question |
Edited by J Torous; submitted 01.Jul.2025; peer-reviewed by K O'Malley, MNA Khalid; comments to author 13.Sep.2025; revised version received 04.Nov.2025; accepted 09.Nov.2025; published 06.Mar.2026.
Copyright©Syeda Umme Salma, Chandra Rekha Renduchintala, Isa Siddique, Evelina Sterling, Sweta Sneha, Nazmus Sakib. Originally published in JMIR Mental Health (https://mental.jmir.org), 06.Mar.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.

